contrastive semantic segmentation learning
Looking Beyond Single Images for Contrastive Semantic Segmentation Learning
We present an approach to contrastive representation learning for semantic segmentation. Our approach leverages the representational power of existing feature extractors to find corresponding regions across images. These cross-image correspondences are used as auxiliary labels to guide the pixel-level selection of positive and negative samples for more effective contrastive learning in semantic segmentation. We show that auxiliary labels can be generated from a variety of feature extractors, ranging from image classification networks that have been trained using unsupervised contrastive learning to segmentation models that have been trained on a small amount of labeled data. We additionally introduce a novel metric for rapidly judging the quality of a given auxiliary-labeling strategy, and empirically analyze various factors that influence the performance of contrastive learning for semantic segmentation. We demonstrate the effectiveness of our method both in the low-data as well as the high-data regime on various datasets. Our experiments show that contrastive learning with our auxiliary-labeling approach consistently boosts semantic segmentation accuracy when compared to standard ImageNet pretraining and outperforms existing approaches of contrastive and semi-supervised semantic segmentation.
Looking Beyond Single Images for Contrastive Semantic Segmentation Learning - Supplementary Material - 1 Additional results 1.1 Controlled experiment on auxiliary label generation
Table 1 reports the results of a controlled experiment evaluating different components in our framework for auxiliary label generation. Positive correspondences are generated by matching pixels across different augmentations of the same image. With respect to the clustering algorithm, K-means performs better than DBSCAN (#4 vs. #5), which is We show qualitative results, comparing different feature extractors in Figure 1. DBSCAN is limited by the memory and computational complexity. Corresponding qualitative results are shown in Figure 3. Tables 3-5 show We observe the best performance when 5% outliers are removed.
- Research Report > Strength High (1.00)
- Research Report > Experimental Study (1.00)
Looking Beyond Single Images for Contrastive Semantic Segmentation Learning
We present an approach to contrastive representation learning for semantic segmentation. Our approach leverages the representational power of existing feature extractors to find corresponding regions across images. These cross-image correspondences are used as auxiliary labels to guide the pixel-level selection of positive and negative samples for more effective contrastive learning in semantic segmentation. We show that auxiliary labels can be generated from a variety of feature extractors, ranging from image classification networks that have been trained using unsupervised contrastive learning to segmentation models that have been trained on a small amount of labeled data. We additionally introduce a novel metric for rapidly judging the quality of a given auxiliary-labeling strategy, and empirically analyze various factors that influence the performance of contrastive learning for semantic segmentation.